595 research outputs found

    Development of an instrument for measuring different types of cognitive load

    Get PDF
    According to cognitive load theory, instructions can impose three types of cognitive load on the learner: intrinsic load, extraneous load, and germane load. Proper measurement of the different types of cognitive load can help us understand why the effectiveness and efficiency of learning environments may differ as a function of instructional formats and learner characteristics. In this article, we present a ten-item instrument for the measurement of the three types of cognitive load. Principal component analysis on data from a lecture in statistics for PhD students (n = 56) in psychology and health sciences revealed a three-component solution, consistent with the types of load that the different items were intended to measure. This solution was confirmed by a confirmatory factor analysis of data from three lectures in statistics for different cohorts of bachelor students in the social and health sciences (ns = 171, 136, and 148), and received further support from a randomized experiment with university freshmen in the health sciences (n = 58)

    LEFT BEHIND: MONITORING THE SOCIAL INCLUSION OF YOUNG AUSTRALIANS WITH SELF- REPORTED LONG TERM HEALTH CONDITIONS, IMPAIRMENTS OR DISABILITIES 2001 - 2009

    Get PDF
    Adolescents and young adults with disabilities are at heightened risk of social exclusion. Exclusion leads to poor outcomes in adulthood which in turn affects individuals’ health and wellbeing and that of their families and society through loss of productive engagement in their communities. Australia’s Social Inclusion Indicators Framework provides indices in domains of participation, resources and multiple and entrenched disadvantage to monitor and report on social inclusion. The Household Income and Labour Dynamics in Australia survey provides data over time on households in Australia. Using these tools we report here on the extent of social inclusion/exclusion of young disabled Australians over the past decade. Relative to their non-disabled peers, young disabled Australians are significantly less likely to do well on participation indicators.Centre for Disability Research and Polic

    A new framework for designing programmes of assessment

    Get PDF
    Research on assessment in medical education has strongly focused on individual measurement instruments and their psychometric quality. Without detracting from the value of this research, such an approach is not sufficient to high quality assessment of competence as a whole. A programmatic approach is advocated which presupposes criteria for designing comprehensive assessment programmes and for assuring their quality. The paucity of research with relevance to programmatic assessment, and especially its development, prompted us to embark on a research project to develop design principles for programmes of assessment. We conducted focus group interviews to explore the experiences and views of nine assessment experts concerning good practices and new ideas about theoretical and practical issues in programmes of assessment. The discussion was analysed, mapping all aspects relevant for design onto a framework, which was iteratively adjusted to fit the data until saturation was reached. The overarching framework for designing programmes of assessment consists of six assessment programme dimensions: Goals, Programme in Action, Support, Documenting, Improving and Accounting. The model described in this paper can help to frame programmes of assessment; it not only provides a common language, but also a comprehensive picture of the dimensions to be covered when formulating design principles. It helps identifying areas concerning assessment in which ample research and development has been done. But, more importantly, it also helps to detect underserved areas. A guiding principle in design of assessment programmes is fitness for purpose. High quality assessment can only be defined in terms of its goals

    Driving lesson or driving test?:A metaphor to help faculty separate feedback from assessment

    Get PDF
    Although there is consensus in the medical education world that feedback is an important and effective tool to support experiential workplace-based learning, learners tend to avoid the feedback associated with direct observation because they perceive it as a high-stakes evaluation with significant consequences for their future. The perceived dominance of the summative assessment paradigm throughout medical education reduces learners’ willingness to seek feedback, and encourages supervisors to mix up feedback with provision of ‘objective’ grades or pass/fail marks. This eye-opener article argues that the provision and reception of effective feedback by clinical supervisors and their learners is dependent on both parties’ awareness of the important distinction between feedback used in coaching towards growth and development (assessment for learning) and reaching a high-stakes judgement on the learner’s competence and fitness for practice (assessment of learning). Using driving lessons and the driving test as a metaphor for feedback and assessment helps supervisors and learners to understand this crucial difference and to act upon it. It is the supervisor’s responsibility to ensure that supervisor and learner achieve a clear mutual understanding of the purpose of each interaction (i.e. feedback or assessment). To allow supervisors to use the driving lesson—driving test metaphor for this purpose in their interactions with learners, it should be included in faculty development initiatives, along with a discussion of the key importance of separating feedback from assessment, to promote a feedback culture of growth and support programmatic assessment of competence
    corecore